专利摘要:
audio input from user. a computing device detects a user viewing the computing device and issues a signal if the user is detected to view the computing device. the computing device receives audio input from the user if the user continues viewing the computing device for a predetermined period of time.
公开号:BR112014018604B1
申请号:R112014018604-9
申请日:2012-04-27
公开日:2022-02-01
发明作者:Evan Scheessele
申请人:Hewlett-Packard Development Company, L.P.;
IPC主号:
专利说明:

FUNDAMENTALS
[001] When accessing a device to input an input such as an audio input, the user may initially use an input component of the computing device, such as a mouse or keyboard, to enable an audio input component of the computing device . The user can continue to use the mouse or keyboard to launch an application on the computing device to receive audio input from the user. As the user manipulates the input components, time can be wasted and the user can make mistakes leading to a frustrating user experience. BRIEF DESCRIPTION OF THE DRAWINGS
[002] Several features and advantages of the described embodiments will be evident from the following detailed description, taken in conjunction with the accompanying drawings, which together illustrate by way of example the features of the disclosed embodiments.
[003] Figure 1 illustrates a computing device with a sensor, an output component and an audio input component according to an example.
[004] Figure 2 illustrates an output component emitting a signal and a sensor detecting a user viewing a computing device according to an example.
[005] Figure 3 illustrates a block diagram of a voice application detecting audio input from a user according to an example.
[006] Figure 4 is a flowchart illustrating a method for detecting audio input according to an example.
[007] Figure 5 is a flowchart illustrating a method for detecting audio input according to an example. DETAILED DESCRIPTION
[008] A computing device includes a sensor, such as an image capture component, for detecting a user viewing the computing device. For the purposes of this order, the user is viewing the computing device if the sensor detects a user's face or eyes looking towards the computing device. If the user is viewing the computing device, an output component such as a visual indicator, an audio output speaker, and/or a haptic feedback engine can provide a cue to the user. The signal can be a visual, audio, and/or haptic feedback signal notifying the user that the computing device is ready to receive audio input from the user. When the signal is emitted, the computing device can use the sensor to detect the user viewing the computing device and/or the output component for a predetermined period of time.
[009] If the computing device determines that the user continues viewing the computing device for the predetermined period of time when the signal is issued, the computing device can confirm that the user would like to interact with the computing device. An audio input component of the computing device, such as a microphone, can then receive audio input from the user. In one embodiment, when receiving audio inputs, the computing device launches a listening mode of the computing device to enable, unmute, and/or calibrate the audio input component to receive user audio input. Audio inputs can be a voice command, a voice query, and/or any additional user-generated sound received by the computing device as an input. As a result, the user can easily interact with the device with audio inputs without physically accessing or manipulating the input components of the computing device.
[0010] Figure 1 illustrates a computing device 100 with a sensor 130, an output component 160, and an audio input component 140 according to an example. In one embodiment, the computing device 100 can be a personal computer, a laptop, a notebook, a tablet, a netbook, an all-in-one system, and/or a server. In another embodiment, the computing device 100 may be a cellular device, a PDA (Personal Digital Assistant), an E (electronic) reader, a tablet, a camera, and/or any supplemental device which may include a sensor 130, an output component 180, and an audio component 140.
[0011] The computing device 100 includes a controller 120, a sensor 130, an output component 180, an audio input component 140, and a communication channel 150 for the computing device 100 and/or one or more components device 100 to communicate with each other. In one embodiment, the computing device 100 further includes a voice application stored on a non-volatile computer-readable medium included in or accessible to the computing device 100. For purposes of this application, the voice application is an application that can be used independently and/or in conjunction with controller 120 to control audio input 145 to computing device 100. For purposes of this application, audio input 145 includes a voice and/or sound from a user received as an input to computing device 100.
[0012] When managing audio input 145, a sensor 130 may initially be used to detect whether a user is viewing the computing device 100. For the purposes of the present application, the sensor 130 is a hardware component, such as a capture component image and/or an infrared component, which can capture a view in front of the computing device 100 to detect a user viewing the computing device 100. Detecting the user viewing the computing device 100 includes determining whether a user's face is toward computing device 100. In one embodiment, controller 320 and/or voice application may use face detection technology and/or eye gaze detection technology to determine whether a user's face or eyes are viewing the computing device 100.
[0013] If the user is detected to be viewing the computing device 100, the controller 120 and/or the voice application instructs an output component 160 to issue a signal 185 to the user. For purposes of the present application, a signal 165 is a signal or notification from the output component 160 to notify the user that the computing device 100 is ready to receive audio input 145 from the user. Output component 180 is a hardware component of computing device 100 that can provide visual cue, audio cue, and/or haptic feedback cue to the user. In one embodiment, the output component 160 can be a visual indicator, an audio output speaker, a haptic motor, and/or any additional component that can emit a signal 165.
[0014] Controller 110 and/or voice application may use sensor 130 to detect whether the user has been viewing computing device 100 for a predetermined period of time. The predetermined period of time may be defined by the controller 120, the voice application, and/or the user. For example, the predetermined period of time can be set to half a second, 2 seconds, or any additional amount of time. If the sensor 130 detects the user continuing to view the computing device 100 for the predetermined amount of time, the controller 110 and/or the voice application determines that the user is viewing the computing device 100 for the predetermined amount of time.
[0015] In another embodiment, upon detecting the user viewing the computing device 100 for the predetermined period of time, the sensor 130 may further detect for the user viewing the output component 180 of the computing device 100. The controller 120 and / or the voice application may identify a user's line of sight to determine whether the user is viewing the computing device 100 and/or the output component 180. The controller 120 and/or the voice application may use face detection and/or eye gaze detection technology to identify the wearer's line of sight.
[0016] If the user continues to view the computing device 100 and/or the output component 160 for the predetermined period of time, the controller 120 and/or the voice application may determine that the user's intention is to interact with the computing device 100 and proceeding to receive audio input 145 from the user with an audio input component 140. For purposes of this application, the audio input component 140 is a hardware component, such as a microphone, which can receive 145 audio inputs from the user.
[0017] For the purposes of this application, an audio input 145 includes user audible sounds that the controller 110 and/or the voice application can receive as input to the computing device 100. For example, the audio input 145 may be a voice command and/or a query to computing device 100. In one embodiment, receiving audio input 145 includes controller 120 and/or voice application launching a listening mode of computing device 100. To For purposes of this application, the listening mode of computing device 100 corresponds to an input mode of computing device 100 for receiving audio inputs 145 from the user. If computing device 100 is in listen mode, component audio input 140 is enabled and muted to receive audio input 145.
[0018] Figure 2 illustrates an output component 260 emitting a signal 285 in response to a sensor 230 detecting a user 205 viewing a computing device 200 according to an example. For purposes of this application, sensor 230 is a hardware component of computing device 200 that detects user 205 viewing computing device 200. Sensor 230 may be an image capture component, an infrared component, and/or or any additional component that can detect the user 205 by viewing the computing device 200. In one embodiment, the sensor 230 can also detect a physical proximity of the user 205 around the computing device 200. The physical proximity of the user 205 corresponds to a position or distance of the user from the audio input component 240. The user 205 can be any person who can interact with the computing device 200 and input one or more audio inputs to the computing device 200.
[0019] Sensor 230 may initially capture a view in front of computing device 200 to detect user 205. If user 205 is detected, controller and/or voice application may utilize face detection technology and/or or eye gaze detection technology to detect whether the user 205 is viewing the computing device 200. The controller and/or voice application may utilize face detection technology and/or eye gaze detection technology to determine if computing device 200 is within line of sight of user 205. If computing device 200 is within line of sight of user, the controller and/or voice application determines that user 205 is viewing the device of computing 200.
[0020] If user 205 is detected to be viewing computing device 200, output component 280 proceeds to issue a signal 265 to notify user 205 that computing device 200 is ready to receive audio input from user 205 In one embodiment, the controller and/or voice application may further determine whether the user continues to view computing device 200 for a predetermined period of time before instructing output component 260 to issue signal 265. As noted above, the predetermined period of time can be set by the controller, the voice application, and/or the user 205 to half a second, 2 seconds, and/or any amount of time.
[0021] If user 205 continues to view computing device 200 for the predetermined period of time, output element 260 may be used to issue a signal 265. For purposes of this order, output component 260 is a component hardware of computing device 200, such as a visual indicator, a display component, an audio output speaker, and/or a haptic engine that outputs the 265 signal. The 265 signal includes at least one of a visual signal. , an audio signal, a haptic feedback, and/or any additional signal 285 that notifies the user 205 that the computing device 200 is ready to receive audio inputs. For example, the visual signal may be an illuminating light source. The audio signal may be a beep being emitted. Haptic feedback may include one or more vibrations.
[0022] In one embodiment, the output component 280 is a directional visual indicator with a limited field of view and the signal 285 is issued to the user 205 based on a position of the user 205. As noted above, the sensor 230 can be used to detect a physical location or proximity to the user 205. The visual directional indicator may include a set of light sources that have a limited field of view. One or more of the light sources within proximity of user 205 may be illuminated to emit signal 285 for user 205 to see. Other light sources outside the proximity of the user 205 are not illuminated. As a result, security can be increased for the user 205 and the computing device 200 by issuing the beacon 265 within a limited field of view of the beacon 265 such that others do not notice the beacon 285 and see the user 205 interacting with it. the computing device 200.
[0023] As signaling 265 is output by output component 260, the controller and/or voice application determines whether user 205 is viewing computing device 200 for the predetermined period of time in association with signaling 265 being issued. For purposes of this application, user 205 can be determined to be viewing computing device 200 in association with signal 265 if user 205 continues to view computing device 200 when signal 265 is issued by output component 160. For example, if user 205 continues to view computing device 200 for the predetermined period of time in parallel with signal 265 being emitted, user 205 is determined to be viewing computing device 200 in association with signal 265 being emitted.
[0024] In another embodiment, as opposed to viewing the computing device 200, the controller and/or the voice application may be detected to the user 205 viewing the output component 260 for the predetermined period of time in association with the signal 265 being issued. In other embodiments, the sensor 230 also detects a change in the facial expression of the user 205 or a continuous gaze of the user 205. Upon detecting a change in the facial expression, the sensor 230 detects a smile, an expression, a wink, a raised eyebrow, and/or any further change in the facial expression of the user 205. Upon detecting the continuous gaze, the sensor 230 detects for the face or eyes of the user 205 continuing to view the computing device 200 or the output component 260 for the period of time predetermined.
[0025] If the user 205 is viewing the computing device 200 and/or the output component 260 for the predetermined period of time in association with the 265 signal being issued, the controller and/or the voice application determines that the user I would like to interact with the 200 computing device with an audio input. As shown in Figure 2, an audio input component 240 can be used to detect user audio input 205. For purposes of this application, the audio input component 240 is a hardware component, such as a microphone, that detects voice, sounds and/or noise as an audio input to computing device 200. In one embodiment, user audio input 205 may be a voice command for computing device 200 to execute an instruction or command. The instruction or command may be to start or end an application, open or close a file, and/or dictate alphanumeric text entry into email, an application, and/or an internet browser.
[0026] In another embodiment, the audio input may be a voice query for information from the computing device 200. The query may be for a state of the computing device 200, for a local weather forecast, and/or or for information retrieved from computing device 200 or other device. During capture of a voice command and/or query from the user 205, the sensor 230 may further capture a view of the user's mouth 205. The view of the user's mouth 205 may be used by the controller and/or the software application. voice to improve the accuracy of identifying a voice command and/or a query from the user 205. In other embodiments, the user 205 may use a hand, body part, and/or accessory to create audio for the user. 240 audio input component to receive as an audio input.
[0027] Figure 3 illustrates a block diagram of a voice application 310 detecting audio input from a user according to an example. As noted above, voice application 310 can be used independently and/or in conjunction with controller 320 to manage audio input. In one embodiment, the voice application 310 may be firmware embedded in one or more of the components of the computing device. In another embodiment, the speech application 310 may be an application accessible from non-volatile computer-readable memory of the computing device. A computer-readable memory is a tangible device that contains, stores, communicates, or transports the application for use by or in connection with the device. In one embodiment, the computer-readable memory is a hard disk drive, compact disk, flash disk, network drive, or any other form of tangible apparatus coupled to the computing device.
[0028] As shown in Figure 3, sensor 330 has detected a user viewing the computing device. In response, controller 320 and/or voice application 310 instructs output component 360 to emit visual, audio, and/or haptic feedback to the user. In one embodiment, when the signal is issued, the controller 320 and/or the voice application 310 can authenticate the user with facial recognition technology. Controller 320 and/or voice application 310 may instruct sensor 330 to capture an image of the user's face and compare the image of the user's face to images of users recognized to the computing device. Recognized user images may be stored locally on a storage component of the computing device or on another device accessible to the 320 controller and/or the 310 voice application.
[0029] In one embodiment, if the image of the user does not match an image of a recognized user, the user fails to authenticate and the controller 320 and/or the voice application 310 disables a listening mode of the computing device. By disabling listening mode, the 320 controller and/or the 310 voice application may disable the 340 audio input component, mute the 340 audio input component, and/or reject any audio input from the user. If the image of the user matches an image of a recognized user, the user has been successfully authenticated and the controller 320 and/or the voice application 310 proceeds to determine whether the user continues to view the computing device or component. output for a predetermined period of time.
[0030] If the user is determined to be viewing the computing device and/or the output component for the predetermined period of time, the controller 320 and/or the input application 310 determines that the user would like to interact with the device of computing and proceeds to launch listening mode. In another embodiment, the controller 320 and/or the voice application 310 may further detect a continuous gaze or a change in the user's facial expression as the user continues to view the computing device and/or the output component. Upon detecting the continuous gaze or change in facial expression, the controller 320 and/or the voice application 310 can further confirm the user's intent to interact with the computing device.
[0031] The 320 controller and/or the 310 voice application can then launch the computing device listening mode. Upon launching listening mode, controller 320 and/or voice application 310 can enable audio input component 340 and unmute audio input component 340 to receive user audio input. In one embodiment, upon launching listening mode, the 320 controller and/or voice application loads a voice profile associated with a user identity. For purposes of this application, the voice profile includes information of the users' voice or speech pattern that the controller 320 and/or the voice application 310 uses to improve detection and recognition of a user's audio input. The information may include a pitch of the user's voice, a rate at which the user speaks, an annunciation of the user's speech, and/or any additional information from the user's voice or speech pattern.
[0032] In one embodiment, the controller 320 and/or the voice application 310 also instruct the sensor 330 to detect a user's physical proximity to the audio input component 340. As noted above, the user's physical proximity corresponds at a user's position or distance from the audio input component 340. Based on the user's position and distance, the controller 320 and/or the voice application 310 may calibrate the audio input component 340 to improve a accuracy of detecting audio inputs from the user. Calibrating the audio input component 340 may include directing the audio input component 340 to receive audio input from a user's position. In another embodiment, calibration of the audio input component 340 may include applying a noise filtering or cancellation algorithm based on the position and/or distance of the user.
[0033] In response to detecting audio input from the user, the controller 320 and/or the voice application 310 identifies a command, instruction, input, and/or query from the computing device associated with the user's audio input . In one embodiment, when audio input is detected by audio input component 340, sensor 330 may also capture a view of the user's mouth. By capturing the view from the user's mouth, an accuracy of identification of commands, instructions and/or computing device queries corresponding to the audio input can be improved. The user mouth view can also be used as a command for the audio input component 340 to start receiving audio input from the user.
[0034] Figure 4 is a flowchart illustrating a method for detecting audio input according to an example. A sensor, as in image capture component, may initially detect a user viewing the computing device at 400. If the user is viewing the computing device, the controller and/or voice application may instruct an output component, as a visual signal, speaker, and/or a haptic motor to emit a visual, audio, and/or haptic feedback signal at 410. The signal can be used to notify the user that the computing device is ready to launch a listening mode to receive audio input from the user. When the signal is issued, the controller and/or voice application can determine if the user continues to view the output element for a predetermined period of time when the signal is issued at 420. If the user continues to view the output component output for the predetermined period of time, the controller and/or the voice application determine that the user would like to input a voice command and/or query and a computing device listening mode is launched to an audio input component receive audio input from the user at 430. The method is then completed. In other embodiments, the method of Figure 4 includes additional steps in addition to and/or instead of those depicted in Figure 4.
[0035] Figure 5 is a flowchart illustrating a method for detecting audio input according to an example. A sensor, such as an image capture component, may be used to capture a user's view and the controller and/or voice application may utilize face detection technology and/or eye gaze detection technology to determine whether the user is viewing the computing device at 500. The controller and/or voice application detects whether the user is viewing the computing device by determining whether a user's face is facing the computing device and/or if an eye gaze of the user is viewing the computing device at 510. If the user is not viewing the computing device, the controller and/or voice application continues to detect the user viewing the computing device at 500.
[0036] If the user is determined to be viewing the computing device, an output component may be used to emit a visual, audio and/or haptic feedback signal to notify the user that the computing device is ready to launch a listening mode at 520. When the signal is issued, the controller and/or voice application can determine whether the user continues to view the output component for a predetermined period of time at 530. If the user does not view the component output for the predetermined period, the controller and/or voice application continues to detect the user viewing the computing device at 500.
[0037] In one embodiment, if the user is determined to be viewing the output component for the predetermined amount of time, the controller and/or voice application can determine if a change in facial expression, such as a raised eyebrow, a wink, and/or a smile, was detected on the user at 540. If no change in facial expression is detected, the controller and/or voice app continue to determine whether the user is viewing the output component for the period predetermined time at 530. If a change in facial expression is detected, the user can then be authenticated with facial recognition technology at 550. If the user fails authentication, a listening mode is not launched and the controller and / or the voice app continues to detect if another user is viewing the computing device at 500.
[0038] If the user is successfully authenticated, the controller and/or voice application loads a voice profile associated with a user identity at 560. The voice profile is used to improve detection and recognition of a user's input. user audio. The controller and/or voice app can then launch the computing device's listening mode, enabling an audio input component, such as a microphone, to receive audio input from the user at 570 In one embodiment, the sensor and/or audio input component may also identify a physical proximity to the user and proceed to calibrate the audio input component based on the user's physical proximity to 580. In another embodiment, with the listening launched, the controller and/or voice app may also disable other forms of input to the computing device, such as touch inputs, keyboard inputs, mouse inputs, and/or other forms of input to the computing device. The method is then completed. In other embodiments, the method of Figure 5 includes additional steps in addition to and/or instead of those depicted in Figure 5.
权利要求:
Claims (11)
[0001]
1. Computing device (100, 200), comprising: a sensor (130, 230, 330) for detecting a user (205) viewing the computing device by detecting a face or eyes of the user facing the computing device ( 100, 200); an output component (160, 260, 360) for issuing a signal (165, 265) if the user (205) is detected to view the computing device (100, 200); and characterized by a controller (120) for launching from a listening mode to enable an audio input component (140, 240, 340) to receive an audio input (145) from the user (205) with the audio input component (140, 240, 340) if the sensor detects the user (205) continuing to view the computing device (100, 200) in association with the signaling (165, 265) being emitted for a predetermined period of time.
[0002]
2. Computing device (100, 200), according to claim 1, characterized in that, as the signal is issued, the controller (120, 320) is configured to authenticate (550) the user (205) with facial recognition technology and disable listening mode if authentication fails.
[0003]
3. Computing device (100, 200) according to claim 1, characterized in that the signaling (165, 265) includes at least one of a visual signal, an audio signal, a haptic feedback indicating to the user (205) that the computing device (100, 200) is ready to launch a listening mode to receive audio inputs.
[0004]
4. Computing device (100, 200) according to claim 1, characterized in that the sensor (130, 230, 330) includes at least one of an image capture component and an infrared component.
[0005]
5. Computing device (100, 200) according to claim 1, characterized in that the output component (160, 260, 360) includes at least one of a visual indicator, an audio speaker, and a haptic motor for signaling (165, 265).
[0006]
6. Computing device (100, 200) according to claim 1, characterized in that the controller (120, 320) detects a continuous gaze from the user (205) or a change in the user's facial expression ( 205) when detecting the user (205) viewing the computing device (100, 200).
[0007]
7. A method for receiving audio input, comprising: detecting (400, 510) a user (205) viewing a computing device (100, 200) with a sensor (130, 230, 330); issuing (410, 520) a signaling (165, 265) with an output component (160, 260, 360) if the user (205) is viewing the computing device (100, 200); determining (420, 530) if the user (205) is viewing the output component (160, 260, 360) for a predetermined period of time; is characterized by launching (430, 570) a listening mode of the computing device (100, 200) to enable an audio input component (140, 240, 340) to receive audio input from the user (205) if the user (205) ) is viewing the output component (160, 260, 360) for the predetermined period of time.
[0008]
8. Method for receiving audio input, according to claim 7, characterized in that it further comprises, as the signal is issued, authenticating (550) the user with facial recognition technology and disabling the listening mode if the authentication fails.
[0009]
9. Method for receiving audio input, according to claim 8, characterized in that it further comprises loading a voice profile (560) associated with a user identity (205).
[0010]
10. Method for receiving audio input, according to claim 7, characterized in that it further comprises identifying a physical proximity of the user (205) to the sensor (130, 230, 330) and calibrating (580) a component of audio input (140, 240, 340) from the computing device (100, 200) based on the physical proximity of the user (205).
[0011]
11. Non-volatile computer readable medium, characterized in that it comprises instructions which, if executed, cause a controller (120, 320) to perform the method as defined in any one of claims 7 to 10.
类似技术:
公开号 | 公开日 | 专利标题
BR112014018604B1|2022-02-01|COMPUTER DEVICE, METHOD FOR RECEIVING AUDIO INPUT AND NON-VOLATILE COMPUTER-READable MEDIUM
EP3179474B1|2019-07-24|User focus activated voice recognition
US9430696B2|2016-08-30|Continuous enrollment for face verification
KR102334272B1|2021-12-01|Trainable sensor-based gesture recognition
US20170185760A1|2017-06-29|Face-Controlled Liveness Verification
KR101971697B1|2019-04-23|Method and apparatus for authenticating user using hybrid biometrics information in a user device
US20150261944A1|2015-09-17|Unobtrusive verification of user identity
TW201626365A|2016-07-16|Speaker identification and unsupervised speaker adaptation techniques
US20120281885A1|2012-11-08|System and method for dynamic facial features for speaker recognition
US10204624B1|2019-02-12|False positive wake word
KR102329765B1|2021-11-23|Method of recognition based on IRIS recognition and Electronic device supporting the same
US20140237589A1|2014-08-21|Personalization Data of an Active Application
US10073671B2|2018-09-11|Detecting noise or object interruption in audio video viewing and altering presentation based thereon
US10269377B2|2019-04-23|Detecting pause in audible input to device
WO2015131550A1|2015-09-11|Method and apparatus for controlling player to enter sleep mode and terminal device
WO2019184946A1|2019-10-03|Human facial recognition control method and mobile terminal
JP2011128766A5|2013-01-31|
CN107885329B|2021-04-16|Rotary camera adjusting method and user terminal
WO2015084927A1|2015-06-11|Controlling connection of input device to electronic devices
US20160294823A1|2016-10-06|Displaying content based on device orientation
US20130215008A1|2013-08-22|Portable electronic device and control method thereof
US20200192485A1|2020-06-18|Gaze-based gesture recognition
TW201636778A|2016-10-16|System and method for controlling operation mode
KR20210011146A|2021-02-01|Apparatus for providing a service based on a non-voice wake-up signal and method thereof
TW201621737A|2016-06-16|Privacy system depending on eyeball recognition and method
同族专利:
公开号 | 公开日
CN104094192A|2014-10-08|
JP6001758B2|2016-10-05|
BR112014018604A8|2021-03-09|
US9626150B2|2017-04-18|
TW201403458A|2014-01-16|
EP2842014A4|2015-12-02|
EP2842014A1|2015-03-04|
CN104094192B|2017-09-29|
US20150033130A1|2015-01-29|
JP2015514254A|2015-05-18|
WO2013162603A1|2013-10-31|
TWI490778B|2015-07-01|
引用文献:
公开号 | 申请日 | 公开日 | 申请人 | 专利标题

JP2004192653A|1997-02-28|2004-07-08|Toshiba Corp|Multi-modal interface device and multi-modal interface method|
US6118888A|1997-02-28|2000-09-12|Kabushiki Kaisha Toshiba|Multi-modal interface apparatus and method|
JPH1124694A|1997-07-04|1999-01-29|Sanyo Electric Co Ltd|Instruction recognition device|
JP2001070293A|1999-09-06|2001-03-21|Toshiba Corp|Radio-diagnostic device|
JP2001134256A|1999-11-05|2001-05-18|Sony Corp|User detecting circuit for display device|
JP2002041084A|2000-07-26|2002-02-08|Victor Co Of Japan Ltd|Interactive speech processing system|
JP2003044089A|2001-07-26|2003-02-14|Matsushita Electric Works Ltd|Device and method for recognizing voice|
GB2388209C|2001-12-20|2005-08-23|Canon Kk|Control apparatus|
US7762665B2|2003-03-21|2010-07-27|Queen's University At Kingston|Method and apparatus for communication between humans and devices|
US7522065B2|2004-10-15|2009-04-21|Microsoft Corporation|Method and apparatus for proximity sensing in a portable electronic device|
US20060192775A1|2005-02-25|2006-08-31|Microsoft Corporation|Using detected visual cues to change computer system operating states|
US7438414B2|2005-07-28|2008-10-21|Outland Research, Llc|Gaze discriminating electronic control apparatus, system, method and computer program product|
JP4992218B2|2005-09-29|2012-08-08|ソニー株式会社|Information processing apparatus and method, and program|
US7429108B2|2005-11-05|2008-09-30|Outland Research, Llc|Gaze-responsive interface to enhance on-screen user reading tasks|
TW200720991A|2005-11-22|2007-06-01|Delta Electronics Inc|Voice control methods|
JP5510951B2|2006-02-01|2014-06-04|トビーテクノロジーアーベー|Generating graphical feedback in computer systems|
CA2717992C|2008-03-12|2018-01-16|E-Lane Systems Inc.|Speech understanding method and system|
WO2010089989A1|2009-02-05|2010-08-12|パナソニック株式会社|Information display device and information display method|
TWI384383B|2009-03-02|2013-02-01|Univ Nat Central|Apparatus and method for recognizing gaze|
US10540976B2|2009-06-05|2020-01-21|Apple Inc.|Contextual voice commands|
US8416715B2|2009-06-15|2013-04-09|Microsoft Corporation|Interest determination for auditory enhancement|
CN101674363A|2009-09-23|2010-03-17|中兴通讯股份有限公司|Mobile equipment and talking method|
US20110205148A1|2010-02-24|2011-08-25|Corriveau Philip J|Facial Tracking Electronic Reader|
US20110310001A1|2010-06-16|2011-12-22|Visteon Global Technologies, Inc|Display reconfiguration based on face/eye tracking|
US9274744B2|2010-09-10|2016-03-01|Amazon Technologies, Inc.|Relative position-inclusive device interfaces|
US9202105B1|2012-01-13|2015-12-01|Amazon Technologies, Inc.|Image analysis for user authentication|US10255566B2|2011-06-03|2019-04-09|Apple Inc.|Generating and processing task items that represent tasks to perform|
CN113470640A|2013-02-07|2021-10-01|苹果公司|Voice trigger of digital assistant|
US20160282515A1|2013-10-21|2016-09-29|Mitsui Chemicals, Inc.|Polymerizable composition for optical material and optical material|
TWI514258B|2014-02-17|2015-12-21|Hooloop Corp|Methods and systems for voice management, and related computer program products|
US10170123B2|2014-05-30|2019-01-01|Apple Inc.|Intelligent assistant for home automation|
WO2015184186A1|2014-05-30|2015-12-03|Apple Inc.|Multi-command single utterance input method|
US9715875B2|2014-05-30|2017-07-25|Apple Inc.|Reducing the need for manual start/end-pointing and trigger phrases|
JP6555858B2|2014-08-01|2019-08-07|シャープ株式会社|Apparatus, audio output method, audio output program, network system, server, and communication apparatus|
KR102337682B1|2014-10-01|2021-12-09|삼성전자주식회사|Display apparatus and Method for controlling thereof|
US9886953B2|2015-03-08|2018-02-06|Apple Inc.|Virtual assistant activation|
KR20160115330A|2015-03-26|2016-10-06|삼성전자주식회사|Method and electronic device for providing content|
FR3034215B1|2015-03-27|2018-06-15|Valeo Comfort And Driving Assistance|CONTROL METHOD, CONTROL DEVICE, SYSTEM AND MOTOR VEHICLE COMPRISING SUCH A CONTROL DEVICE|
TWI564791B|2015-05-19|2017-01-01|卡訊電子股份有限公司|Broadcast control system, method, computer program product and computer readable medium|
US10200824B2|2015-05-27|2019-02-05|Apple Inc.|Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device|
JP6516585B2|2015-06-24|2019-05-22|パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America|Control device, method thereof and program|
US10747498B2|2015-09-08|2020-08-18|Apple Inc.|Zero latency digital assistant|
US10586535B2|2016-06-10|2020-03-10|Apple Inc.|Intelligent digital assistant in a multi-tasking environment|
DK201670540A1|2016-06-11|2018-01-08|Apple Inc|Application integration with a digital assistant|
US10103699B2|2016-09-30|2018-10-16|LenovoPte. Ltd.|Automatically adjusting a volume of a speaker of a device based on an amplitude of voice input to the device|
US10366691B2|2017-07-11|2019-07-30|Samsung Electronics Co., Ltd.|System and method for voice command context|
US20190050195A1|2017-08-14|2019-02-14|LenovoPte. Ltd.|Output provision based on gaze detection|
WO2019212567A1|2018-05-04|2019-11-07|Google Llc|Invoking automated assistant function based on detected gesture and gaze|
EP3743794A2|2018-05-04|2020-12-02|Google LLC|Hot-word free adaptation of automated assistant function|
US10928918B2|2018-05-07|2021-02-23|Apple Inc.|Raise to speak|
DK180639B1|2018-06-01|2021-11-04|Apple Inc|DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT|
DK179822B1|2018-06-01|2019-07-12|Apple Inc.|Voice interaction at a primary device to access call functionality of a companion device|
CN108854052A|2018-06-22|2018-11-23|福州市马尾区朱雀网络信息技术有限公司|A kind of game sound real-time feedback method and system|
CN112639719A|2018-09-07|2021-04-09|索尼公司|Terminal device, control method of terminal device, and storage medium|
US11151993B2|2018-12-28|2021-10-19|Baidu Usa Llc|Activating voice commands of a smart display device based on a vision-based mechanism|
US11271762B2|2019-05-10|2022-03-08|Citrix Systems, Inc.|Systems and methods for virtual meetings|
DK180129B1|2019-05-31|2020-06-02|Apple Inc.|User activity shortcut suggestions|
CN110602556A|2019-09-20|2019-12-20|深圳创维-Rgb电子有限公司|Playing method, cloud server and storage medium|
法律状态:
2018-12-11| B06F| Objections, documents and/or translations needed after an examination request according [chapter 6.6 patent gazette]|
2020-04-14| B06U| Preliminary requirement: requests with searches performed by other patent offices: procedure suspended [chapter 6.21 patent gazette]|
2021-09-08| B06A| Patent application procedure suspended [chapter 6.1 patent gazette]|
2021-12-28| B09A| Decision: intention to grant [chapter 9.1 patent gazette]|
2022-02-01| B16A| Patent or certificate of addition of invention granted [chapter 16.1 patent gazette]|Free format text: PRAZO DE VALIDADE: 20 (VINTE) ANOS CONTADOS A PARTIR DE 27/04/2012, OBSERVADAS AS CONDICOES LEGAIS. |
优先权:
申请号 | 申请日 | 专利标题
PCT/US2012/035538|WO2013162603A1|2012-04-27|2012-04-27|Audio input from user|
[返回顶部]